33 research outputs found

    The Impact of Word, Multiple Word, and Sentence Input on Virtual Keyboard Decoding Performance

    Get PDF
    Entering text on non-desktop computing devices is often done via an onscreen virtual keyboard. Input on such keyboards normally consists of a sequence of noisy tap events that specify some amount of text, most commonly a single word. But is single word-at-a-time entry the best choice? This paper compares user performance and recognition accuracy of wordat- a-time, phrase-at-a-time, and sentence-at-a-time text entry on a smartwatch keyboard. We evaluate the impact of differing amounts of input in both text copy and free composition tasks. We found providing input of an entire sentence significantly improved entry rates from 26wpm to 32wpm while keeping character error rates below 4%. In offline experiments with more processing power and memory, sentence input was recognized with a much lower 2.0% error rate. Our findings suggest virtual keyboards can enhance performance by encouraging users to provide more input per recognition event.This work was supported by Google Faculty awards (K.V. and P.O.K.

    Statistical Language Modelling

    Get PDF
    Grammar-based natural language processing has reached a level where it can `understand' language to a limited degree in restricted domains. For example, it is possible to parse textual material very accurately and assign semantic relations to parts of sentences. An alternative approach originates from the work of Shannon over half a century ago [41], [42]. This approach assigns probabilities to linguistic events, where mathematical models are used to represent statistical knowledge. Once models are built, we decide which event is more likely than the others according to their probabilities. Although statistical methods currently use a very impoverished representation of speech and language (typically finite state), it is possible to train the underlying models from large amounts of data. Importantly, such statistical approaches often produce useful results. Statistical approaches seem especially well-suited to spoken language which is often spontaneous or conversational and not readily amenable to standard grammar-based approaches

    Retrieval and browsing of spoken content

    No full text

    Investigating the Global Semantic Impact of Speech Recognition Error on Spoken Content Collections

    No full text

    Are Morphosyntactic Taggers Suitable to Improve Automatic Transcription?

    Get PDF
    International audienceThe aim of our paper is to study the interest of part of speech (POS) tagging to improve speech recognition. We first evaluate the part of misrecognized words that can be corrected using POS information; the analysis of a short extract of French radio broadcast news shows that an absolute decrease of the word error rate by 1.1% can be expected. We also demonstrate quantitatively that traditional POS taggers are reliable when applied to spoken corpus, including automatic transcriptions. This new result enables us to effectively use POS tag knowledge to improve, in a postprocessing stage, the quality of transcriptions, especially correcting agreement errors

    Large Scale Distributed Acoustic Modeling With Back-Off N{\rm N}-Grams

    No full text

    Phoneme-Lattice to Phoneme-Sequence Matching Algorithm Based on Dynamic Programming

    No full text

    Towards a Dynamic Syntax for Language Modelling

    No full text
    corecore